Domain Adaptation of a Dependency Parser with a Class-Class Selectional Preference Model

نویسندگان

  • Raphael Cohen
  • Yoav Goldberg
  • Michael Elhadad
چکیده

When porting parsers to a new domain, many of the errors are related to wrong attachment of out-of-vocabulary words. Since there is no available annotated data to learn the attachment preferences of the target domain words, we attack this problem using a model of selectional preferences based on domainspecific word classes. Our method uses Latent Dirichlet Allocations (LDA) to learn a domain-specific Selectional Preference model in the target domain using un-annotated data. The model provides features that model the affinities among pairs of words in the domain. To incorporate these new features in the parsing model, we adopt the co-training approach and retrain the parser with the selectional preferences features. We apply this method for adapting Easy First, a fast nondirectional parser trained on WSJ, to the biomedical domain (Genia Treebank). The Selectional Preference features reduce error by 4.5% over the co-training baseline.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Exploiting Web-Derived Selectional Preference to Improve Statistical Dependency Parsing

In this paper, we present a novel approach which incorporates the web-derived selectional preferences to improve statistical dependency parsing. Conventional selectional preference learning methods have usually focused on word-to-class relations, e.g., a verb selects as its subject a given nominal class. This paper extends previous work to wordto-word selectional preferences by using webscale d...

متن کامل

Hybrid Ways to Improve Domain Independence in an ML Dependency Parser

This paper reports a hybridization experi­ ment, where a baseline ML dependency pars­ er, LingPars, was allowed access to Con­ straint Grammar analyses provided by a rule­ based parser (EngGram) for the same data. Descriptive compatibility issues and their in­ fluence on performance are discussed. The hybrid system performed considerably better than its ML baseline, ...

متن کامل

Cmpt 825: Natural Language Processing 1.1 Hiding a Semantic Hierarchy in a Markov Model [1] 1.1.1 General Concepts

We know that in logic a predicate is a relation between its arguments. In other words, a predicate defines constraints between its arguments. A predicate ρ(v, r, c) is called selectional restriction where v is a verb, r is a role or an object and c is called a class which is a noun. Selectional preference σ : (v, r, c) → a is a function of these predicates to a real number. Where a shows the de...

متن کامل

Learning class-to-class selectional preferences

Selectional preference learning methods have usually focused on wordto-class relations, e.g., a verb selects as its subject a given nominal class. This papers extends previous statistical models to class-to-class preferences, and presents a model that learns selectional preferences for classes of verbs. The motivation is twofold: different senses of a verb may have different preferences, and so...

متن کامل

Integrating selectional preferences in WordNet

Selectional preference learning methods have usually focused on word-to-class relations, e.g., a verb selects as its subject a given nominal class. This paper extends previous statistical models to class-to-class preferences, and presents a model that learns selectional preferences for classes of verbs, together with an algorithm to integrate the learned preferences in WordNet. The theoretical ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2012